Bayesian Out-Trees
نویسنده
چکیده
A Bayesian treatment of latent directed graph structure for non-iid data is provided where each child datum is sampled with a directed conditional dependence on a single unknown parent datum. The latent graph structure is assumed to lie in the family of directed out-tree graphs which leads to efficient Bayesian inference. The latent likelihood of the data and its gradients are computable in closed form via Tutte’s directed matrix tree theorem using determinants and inverses of the out-Laplacian. This novel likelihood subsumes iid likelihood, is exchangeable and yields efficient unsupervised and semi-supervised learning algorithms. In addition to handling taxonomy and phylogenetic datasets the out-tree assumption performs surprisingly well as a semi-parametric density estimator on standard iid datasets. Experiments with unsupervised and semisupervised learning are shown on various UCI and taxonomy datasets.
منابع مشابه
Learning Bayesian Network Using Parse Trees for Extraction of Protein-Protein Interaction
Extraction of protein-protein interactions from scientific papers is a relevant task in the biomedical field. Machine learning-based methods such as kernel-based represent the state-of-the-art in this task. Many efforts have focused on obtaining new types of kernels in order to employ syntactic information, such as parse trees, to extract interactions from sentences. These methods have reached ...
متن کاملDynamic importance sampling in Bayesian networks using factorisation of probability trees
Factorisation of probability trees is a useful tool for inference in Bayesian networks. Probabilistic potentials some of whose parts are proportional can be decomposed as a product of smaller trees. Some algorithms, like lazy propagation, can take advantage of this fact. Also, the factorisation can be used as a tool for approximating inference, if the decomposition is carried out even if the pr...
متن کاملPrediction with Missing Data via Bayesian Additive Regression Trees
We present a method for incorporating missing data into general forecasting problems which use non-parametric statistical learning. We focus on a tree-based method, Bayesian Additive Regression Trees (BART), enhanced with “Missingness Incorporated in Attributes,” an approach recently proposed for incorporating missingness into decision trees. This procedure extends the native partitioning mecha...
متن کاملPaper Learning Bayesian Belief Networks Based on the Minimum Description Length Principle: Basic Properties
SUMMARY This paper addresses the problem of learning Bayesian belief networks (BBN) based on the minimum description length (MDL) principle. First, we give a formula of description length based on which the MDL-based procedure learns a BBN. Secondly, we point out that the diierence between the MDL-based and Cooper and Herskovits procedures is essentially in the priors rather than in the approac...
متن کاملIncremental Thin Junction Trees for Dynamic Bayesian Networks
We present Incremental Thin Junction Trees, a general framework for approximate inference in static and dynamic Bayesian Networks. This framework incrementally builds junction trees representing probability distributions over a dynamically changing set of variables. Variables and their conditional probability tables can be introduced into the junction tree Υ, they can be summed out of Υ and Υ c...
متن کاملA. W. F. Edwards and the Origin of Bayesian Phylogenetics
In the early 1960s, Anthony Edwards and Luca Cavalli-Sforza made an effort to apply R. A. Fisher’s maximum-likelihood method to estimate genealogical trees of human populations using gene-frequency data. They used the Yule branching process to describe the probabilities of the trees and branching times and the Brownian-motion process to model the drift of gene frequencies (after a suitable tran...
متن کامل